Quadratic Form (statistics)
   HOME

TheInfoList



OR:

In
multivariate statistics Multivariate statistics is a subdivision of statistics encompassing the simultaneous observation and analysis of more than one outcome variable, i.e., '' multivariate random variables''. Multivariate statistics concerns understanding the differ ...
, if \varepsilon is a
vector Vector most often refers to: * Euclidean vector, a quantity with a magnitude and a direction * Disease vector, an agent that carries and transmits an infectious pathogen into another living organism Vector may also refer to: Mathematics a ...
of n
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s, and \Lambda is an n-dimensional
symmetric matrix In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, Because equal matrices have equal dimensions, only square matrices can be symmetric. The entries of a symmetric matrix are symmetric with ...
, then the scalar quantity \varepsilon^T\Lambda\varepsilon is known as a quadratic form in \varepsilon.


Expectation

It can be shown that :\operatorname\left varepsilon^T\Lambda\varepsilon\right\operatorname\left Lambda \Sigma\right+ \mu^T\Lambda\mu where \mu and \Sigma are the
expected value In probability theory, the expected value (also called expectation, expectancy, expectation operator, mathematical expectation, mean, expectation value, or first Moment (mathematics), moment) is a generalization of the weighted average. Informa ...
and variance-covariance matrix of \varepsilon, respectively, and tr denotes the trace of a matrix. This result only depends on the existence of \mu and \Sigma; in particular, normality of \varepsilon is ''not'' required. A book treatment of the topic of quadratic forms in random variables is that of Mathai and Provost.


Proof

Since the quadratic form is a scalar quantity, \varepsilon^T\Lambda\varepsilon = \operatorname(\varepsilon^T\Lambda\varepsilon). Next, by the cyclic property of the trace operator, : \operatorname operatorname(\varepsilon^T\Lambda\varepsilon)= \operatorname operatorname(\Lambda\varepsilon\varepsilon^T) Since the trace operator is a
linear combination In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of the components of the matrix, it therefore follows from the linearity of the expectation operator that : \operatorname operatorname(\Lambda\varepsilon\varepsilon^T)= \operatorname(\Lambda \operatorname(\varepsilon\varepsilon^T)). A standard property of variances then tells us that this is : \operatorname(\Lambda (\Sigma + \mu \mu^T)). Applying the cyclic property of the trace operator again, we get : \operatorname(\Lambda\Sigma) + \operatorname(\Lambda \mu \mu^T) = \operatorname(\Lambda\Sigma) + \operatorname(\mu^T\Lambda\mu) = \operatorname(\Lambda\Sigma) + \mu^T\Lambda\mu.


Variance in the Gaussian case

In general, the variance of a quadratic form depends greatly on the distribution of \varepsilon. However, if \varepsilon ''does'' follow a multivariate normal distribution, the variance of the quadratic form becomes particularly tractable. Assume for the moment that \Lambda is a symmetric matrix. Then, :\operatorname \left varepsilon^T\Lambda\varepsilon\right= 2\operatorname\left Lambda \Sigma\Lambda \Sigma\right+ 4\mu^T\Lambda\Sigma\Lambda\mu. In fact, this can be generalized to find the
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
between two quadratic forms on the same \varepsilon (once again, \Lambda_1 and \Lambda_2 must both be symmetric): :\operatorname\left varepsilon^T\Lambda_1\varepsilon,\varepsilon^T\Lambda_2\varepsilon\right2\operatorname\left Lambda _1\Sigma\Lambda_2 \Sigma\right+ 4\mu^T\Lambda_1\Sigma\Lambda_2\mu. In addition, a quadratic form such as this follows a generalized chi-squared distribution.


Computing the variance in the non-symmetric case

The case for general \Lambda can be derived by noting that :\varepsilon^T\Lambda^T\varepsilon=\varepsilon^T\Lambda\varepsilon so :\varepsilon^T\tilde\varepsilon=\varepsilon^T\left(\Lambda+\Lambda^T\right)\varepsilon/2 is a quadratic form in the symmetric matrix \tilde=\left(\Lambda+\Lambda^T\right)/2, so the mean and variance expressions are the same, provided \Lambda is replaced by \tilde therein.


Examples of quadratic forms

In the setting where one has a set of observations y and an operator matrix H, then the residual sum of squares can be written as a quadratic form in y: :\textrm=y^T(I-H)^T (I-H)y. For procedures where the matrix H is
symmetric Symmetry () in everyday life refers to a sense of harmonious and beautiful proportion and balance. In mathematics, the term has a more precise definition and is usually used to refer to an object that is invariant under some transformations ...
and
idempotent Idempotence (, ) is the property of certain operations in mathematics and computer science whereby they can be applied multiple times without changing the result beyond the initial application. The concept of idempotence arises in a number of pl ...
, and the errors are Gaussian with covariance matrix \sigma^2I, \textrm/\sigma^2 has a
chi-squared distribution In probability theory and statistics, the \chi^2-distribution with k Degrees of freedom (statistics), degrees of freedom is the distribution of a sum of the squares of k Independence (probability theory), independent standard normal random vari ...
with k degrees of freedom and noncentrality parameter \lambda, where :k=\operatorname\left I-H)^T(I-H)\right/math> :\lambda=\mu^T(I-H)^T(I-H)\mu/2 may be found by matching the first two
central moment In probability theory and statistics, a central moment is a moment of a probability distribution of a random variable about the random variable's mean; that is, it is the expected value of a specified integer power of the deviation of the random ...
s of a noncentral chi-squared random variable to the expressions given in the first two sections. If Hy estimates \mu with no
bias Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is inaccurate, closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individ ...
, then the noncentrality \lambda is zero and \textrm/\sigma^2 follows a central chi-squared distribution.


See also

*
Quadratic form In mathematics, a quadratic form is a polynomial with terms all of degree two (" form" is another name for a homogeneous polynomial). For example, 4x^2 + 2xy - 3y^2 is a quadratic form in the variables and . The coefficients usually belong t ...
*
Covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
* Matrix representation of conic sections


References

{{Reflist Statistical theory
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...